Adaptive hybrid activation function for deep neural networks
نویسندگان
چکیده
The adaptive hybrid activation function (AHAF) is proposed that combines the properties of rectifier units and squashing functions. can be used as a drop-in replacement for ReLU, SiL Swish activations deep neural networks evolve to one such functions during training. effectiveness was evaluated on image classification task using Fashion-MNIST CIFAR-10 datasets. evaluation shows with AHAF achieve better accuracy comparing their base implementations use ReLU SiL. A double-stage parameter tuning process training proposed. approach sufficiently simple from implementation standpoint provides high performance network process.
منابع مشابه
Adaptive Activation Functions for Deep Networks
Artificial neural networks loosely mimic the complex web of nearly 100 trillion connections in the human brain. Deep neural networks, and specifically convolutional neural networks, have recently demonstrated breakthrough performances in the pattern recognition community. Studies on the network depth, regularization, filters, choice of activation function, and training parameters are numerous. ...
متن کاملAdaptive dropout for training deep neural networks
Recently, it was shown that deep neural networks can perform very well if the activities of hidden units are regularized during learning, e.g, by randomly dropping out 50% of their activities. We describe a method called ‘standout’ in which a binary belief network is overlaid on a neural network and is used to regularize of its hidden units by selectively setting activities to zero. This ‘adapt...
متن کاملNonparametric regression using deep neural networks with ReLU activation function
Consider the multivariate nonparametric regression model. It is shown that estimators based on sparsely connected deep neural networks with ReLU activation function and properly chosen network architecture achieve the minimax rates of convergence (up to log n-factors) under a general composition assumption on the regression function. The framework includes many well-studied structural constrain...
متن کاملActivation Ensembles for Deep Neural Networks
Many activation functions have been proposed in the past, but selecting an adequate one requires trial and error. We propose a new methodology of designing activation functions within a neural network at each layer. We call this technique an “activation ensemble” because it allows the use of multiple activation functions at each layer. This is done by introducing additional variables, α, at eac...
متن کاملWhy Deep Neural Networks for Function Approximation?
Recently there has been much interest in understanding why deep neural networks are preferred to shallow networks. We show that, for a large class of piecewise smooth functions, the number of neurons needed by a shallow network to approximate a function is exponentially larger than the corresponding number of neurons needed by a deep network for a given degree of function approximation. First, ...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Sistemnì doslìdžennâ ta ìnformacìjnì tehnologìï
سال: 2022
ISSN: ['1681-6048', '2308-8893']
DOI: https://doi.org/10.20535/srit.2308-8893.2022.1.07